A Subspace Learning Approach for High Dimensional Matrix Decomposition with Efficient Column/Row Sampling
نویسندگان
چکیده
This paper presents a new randomized approach to high-dimensional low rank (LR) plus sparse matrix decomposition. For a data matrix D ∈ RN1×N2 , the complexity of conventional decomposition methods is O(N1N2r), which limits their usefulness in big data settings (r is the rank of the LR component). In addition, the existing randomized approaches rely for the most part on uniform random sampling, which may be inefficient for many real world data matrices. The proposed subspace learning-based approach recovers the LR component using only a small subset of the columns/rows of data and reduces complexity to O(max(N1, N2)r). Even when the columns/rows are sampled uniformly at random, the sufficient number of sampled columns/rows is shown to be roughly O(rμ), where μ is the coherency parameter of the LR component. In addition, efficient sampling algorithms are proposed to address the problem of column/row sampling from structured data.
منابع مشابه
Tensor Subspace Analysis
Previous work has demonstrated that the image variations of many objects (human faces in particular) under variable lighting can be effectively modeled by low dimensional linear spaces. The typical linear subspace learning algorithms include Principal Component Analysis (PCA), Linear Discriminant Analysis (LDA), and Locality Preserving Projection (LPP). All of these methods consider an n1 × n2 ...
متن کاملProbabilistic Rank-One Matrix Analysis with Concurrent Regularization
As a classical subspace learning method, Probabilistic PCA (PPCA) has been extended to several bilinear variants for dealing with matrix observations. However, they are all based on the Tucker model, leading to a restricted subspace representation and the problem of rotational ambiguity. To address these problems, this paper proposes a bilinear PPCA method named as Probabilistic Rank-One Matrix...
متن کاملSubspace Sampling and Relative-Error Matrix Approximation: Column-Row-Based Methods
Much recent work in the theoretical computer science, linear algebra, and machine learning has considered matrix decompositions of the following form: given an m×n matrix A, decompose it as a product of three matrices, C, U , and R, where C consists of a small number of columns of A, R consists of a small number of rows of A, and U is a small carefully constructed matrix that guarantees that th...
متن کاملEla Accurate and Efficient Ldu Decompositions of Diagonally Dominant M-matrices
An efficient method for the computation to high relative accuracy of the LDU decomposition of an n × n row diagonally dominant M–matrix is presented, assuming that the off–diagonal entries and row sums are given. This method costs an additional O(n) elementary operations over the cost of Gaussian elimination, and leads to a lower triangular, column diagonally dominant matrix and an upper triang...
متن کاملImage Classification via Sparse Representation and Subspace Alignment
Image representation is a crucial problem in image processing where there exist many low-level representations of image, i.e., SIFT, HOG and so on. But there is a missing link across low-level and high-level semantic representations. In fact, traditional machine learning approaches, e.g., non-negative matrix factorization, sparse representation and principle component analysis are employed to d...
متن کامل